Goto

Collaborating Authors

 un-participated client


Appendices for the Paper pFL Bench A Comprehensive Benchmark for Personalized Federated Learning

Neural Information Processing Systems

Sec.A: the details of adopted datasets and models ( e.g., tasks, heterogeneous partitions, and FL processes as shown in Figure 4 in the main body of the paper. The CIF AR10 is a popular dataset for 10-class image classification containing 60,000 colored images with resolution of 32x32 pixels. The train/valid/test sets are with a ratio of about 3:1:5. The train/valid/test sets are with ratio about 14:3:3. For all these experimental datasets, we randomly select 20% clients as new clients that do not participate in the FL processes.


pFL-Bench: A Comprehensive Benchmark for Personalized Federated Learning

Chen, Daoyuan, Gao, Dawei, Kuang, Weirui, Li, Yaliang, Ding, Bolin

arXiv.org Artificial Intelligence

Personalized Federated Learning (pFL), which utilizes and deploys distinct local models, has gained increasing attention in recent years due to its success in handling the statistical heterogeneity of FL clients. However, standardized evaluation and systematical analysis of diverse pFL methods remain a challenge. Firstly, the highly varied datasets, FL simulation settings and pFL implementations prevent easy and fair comparisons of pFL methods. Secondly, the current pFL literature diverges in the adopted evaluation and ablation protocols. Finally, the effectiveness and robustness of pFL methods are under-explored in various practical scenarios, such as the generalization to new clients and the participation of resource-limited clients. To tackle these challenges, we propose the first comprehensive pFL benchmark, pFL-Bench, for facilitating rapid, reproducible, standardized and thorough pFL evaluation. The proposed benchmark contains more than 10 dataset variants in various application domains with a unified data partition and realistic heterogeneous settings; a modularized and easy-to-extend pFL codebase with more than 20 competitive pFL method implementations; and systematic evaluations under containerized environments in terms of generalization, fairness, system overhead, and convergence. We highlight the benefits and potential of state-of-the-art pFL methods and hope the pFL-Bench enables further pFL research and broad applications that would otherwise be difficult owing to the absence of a dedicated benchmark.